Constructive Neural Networks with Regularization
نویسندگان
چکیده
In this paper we present a regularization approach to the training of all the network weights in cascadecorrelation type constructive neural networks. Especially, the case of regularizing the output neuron of the network is presented. In this case, the output weights are trained by employing a regularized objective function containing a penalty term which is proportional to the weight values of the unit being trained. It is shown that the training can still be done with the pseudo-inverse method of linear regression if the output unit employs linear activation function. The degree of regularization and the smoothness of network mapping can be adjusted by changing the value of the regularization parameter. The investigated algorithms were Cascade-Correlation, Modified Cascade-Correlation, Cascade, and Fixed Cascade Error. The algorithms having the regularization of the hidden and output units were compared with the ones having only the regularization of the hidden units and with those having no regularization at all. The simulation results show that the regularization of the output unit is highly beneficial. It leads to better generalization performance and in many cases to lower computational costs when compared to the partially and non-regulated versions of the same algorithms. Key-Words: Constructive neural networks, regularization, generalization, cascade-correlation, classification, regression.
منابع مشابه
Bayesian Regularization in Constructive Neural Networks
In this paper, we study the incorporation of Bayesian reg-ularization into constructive neural networks. The degree of regulariza-tion is automatically controlled in the Bayesian inference framework and hence does not require manual setting. Simulation shows that regular-ization, with input training using a full Bayesian approach, produces networks with better generalization performance and low...
متن کاملTransfer functions: hidden possibilities for better neural networks
Sigmoidal or radial transfer functions do not guarantee the best generalization nor fast learning of neural networks. Families of parameterized transfer functions provide flexible decision borders. Networks based on such transfer functions should be small and accurate. Several possibilities of using transfer functions of different types in neural models are discussed, including enhancement of i...
متن کاملSupervised Models C1.2 Multilayer perceptrons
This section introduces multilayer perceptrons, which are the most commonly used type of neural network. The popular backpropagation training algorithm is studied in detail. The momentum and adaptive step size techniques, which are used for accelerated training, are discussed. Other acceleration techniques are briefly referenced. Several implementation issues are then examined. The issue of gen...
متن کاملPredictive Abilities of Bayesian Regularization and Levenberg–Marquardt Algorithms in Artificial Neural Networks: A Comparative Empirical Study on Social Data
The objective of this study is to compare the predictive ability of Bayesian regularization with Levenberg–Marquardt Artificial Neural Networks. To examine the best architecture of neural networks, the model was tested with one-, two-, three-, four-, and five-neuron architectures, respectively. MATLAB (2011a) was used for analyzing the Bayesian regularization and Levenberg–Marquardt learning al...
متن کاملRegularization Parameter Selection for Faulty Neural Networks
Regularization techniques have attracted many researches in the past decades. Most focus on designing the regularization term, and few on the optimal regularization parameter selection, especially for faulty neural networks. As is known that in the real world, the node faults often inevitably take place, which would lead to many faulty network patterns. If employing the conventional method, i.e...
متن کامل